NOVEL RESAMPLING METHODS FOR TUNING PARAMETER SELECTION IN ROBUST SPARSE REGRESSION MODELING by

نویسندگان

  • Heewon Park
  • H. Park
چکیده

The robust lasso-type regularized regression is a useful tool for simultaneous estimation and variable selection even in the presence of outliers. Crucial issues in the robust modeling procedure include the selection of regularization parameters and also a tuning constant in outlier detection. Although the performance of the robust sparse regression strongly depends on the proper choice of these tuning parameters, little attention was paid for this issue, particularly in the presence of outliers. We consider the problem of choosing the tuning parameters and propose an information-theoretic criterion based on the bootstrap. Although the bootstrap information criterion has several advantages on its flexibility and weak assumptions, a bootstrap sample may contain more outliers compared with those included in the original sample, since the bootstrap sample is drawn randomly. This implies that the bootstrap information criterion may be obtained from the highly contaminated bootstrap sample by outliers, so the resulting criterion may produce biased results. In order to overcome the drawback, we propose a robust bootstrap information criterion via winsorizing technique (Srivastava et al., 2010) in line with the efficient bootstrap information criterion (Konishi and Kitagawa, 1996) for choosing an optimal set of tuning parameters. Monte Carlo simulations and real data analysis are conducted to investigate the effectiveness of the proposed method. We observe that the proposed robust efficient bootstrap information criterion produces reliable model estimates and performs well in the presence of outliers.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

A Novel Resampling Method for Variable Selection in Robust Regression

Variable selection in regression analysis is of vital importance for data analyst and researcher to fit the parsimonious regression model. With the inundation of large number of predictor variables and large data sets requiring analysis and empirical modeling, contamination becomes usual problem. Accordingly, robust regression estimators are designed to easily fit contaminated data sets. In the...

متن کامل

Stock Price Prediction using Machine Learning and Swarm Intelligence

Background and Objectives: Stock price prediction has become one of the interesting and also challenging topics for researchers in the past few years. Due to the non-linear nature of the time-series data of the stock prices, mathematical modeling approaches usually fail to yield acceptable results. Therefore, machine learning methods can be a promising solution to this problem. Methods: In this...

متن کامل

Robust Estimation in Linear Regression with Molticollinearity and Sparse Models

‎One of the factors affecting the statistical analysis of the data is the presence of outliers‎. ‎The methods which are not affected by the outliers are called robust methods‎. ‎Robust regression methods are robust estimation methods of regression model parameters in the presence of outliers‎. ‎Besides outliers‎, ‎the linear dependency of regressor variables‎, ‎which is called multicollinearity...

متن کامل

Path Thresholding: Asymptotically Tuning-Free High-Dimensional Sparse Regression

In this paper, we address the challenging problem of selecting tuning parameters for high-dimensional sparse regression. We propose a simple and computationally efficient method, called path thresholding (PaTh), that transforms any tuning parameter-dependent sparse regression algorithm into an asymptotically tuning-free sparse regression algorithm. More specifically, we prove that, as the probl...

متن کامل

Efficient and Robust Parameter Tuning for Heuristic Algorithms

The main advantage of heuristic or metaheuristic algorithms compared to exact optimization methods is their ability in handling large-scale instances within a reasonable time, albeit at the expense of losing a guarantee for achieving the optimal solution. Therefore, metaheuristic techniques are appropriate choices for solving NP-hard problems to near optimality. Since the parameters of heuristi...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2012